Goto

Collaborating Authors

 classification rule


On Binary Classification in Extreme Regions

Neural Information Processing Systems

In pattern recognition, a random label Y is to be predicted based upon observing a random vector X valued in $\mathbb{R}^d$ with d> 1 by means of a classification rule with minimum probability of error.




6fee03d84375a159ecd3769ebbacae83-Supplemental-Conference.pdf

Neural Information Processing Systems

Convergence of stochastic gradient descent for non-smooth problems is a known result. For completeness, wereproduce and adapt ausual proof toour setting. Let us denote byF the class of functions fromX toY we are going to work with. Assumption 1 states that we have a well-specified modelF to estimate the median,i.e. Let us begin by controlling the estimation error.



Strong Memory, Weak Control: An Empirical Study of Executive Functioning in LLMs

de Langis, Karin, Park, Jong Inn, Hu, Bin, Le, Khanh Chi, Schramm, Andreas, Mensink, Michael C., Elfenbein, Andrew, Kang, Dongyeop

arXiv.org Artificial Intelligence

Working memory, or the ability to hold and manipulate information in the mind, is a critical component of human intelligence and executive functioning. It is correlated with performance on various cognitive tasks, including measures of fluid intelligence, which encompasses reasoning and problem solving. We use a comprehensive set of classic working memory tasks to estimate the working memory capacity of large language models (LLMs). We find that in most cases, LLMs exceed normative human scores. However, we do not find that the increased capacity of working memory is associated with higher performance on other executive functioning tasks or problem solving benchmarks. These results suggest that LLMs may have deficits in attentional control and cognitive flexibility, which result in difficulties with inhibiting automatic responses and adapting to shifting information. Our findings suggest that current reasoning models have mixed results in compensating for these deficits.


On Binary Classification in Extreme Regions

Neural Information Processing Systems

In pattern recognition, a random label Y is to be predicted based upon observing a random vector X valued in $\mathbb{R}^d$ with d> 1 by means of a classification rule with minimum probability of error.


Leveraging Association Rules for Better Predictions and Better Explanations

Audemard, Gilles, Coste-Marquis, Sylvie, Marquis, Pierre, Sabiri, Mehdi, Szczepanski, Nicolas

arXiv.org Artificial Intelligence

We present a new approach to classification that combines data and knowledge. In this approach, data mining is used to derive association rules (possibly with negations) from data. Those rules are leveraged to increase the predictive performance of tree-based models (decision trees and random forests) used for a classification task. They are also used to improve the corresponding explanation task through the generation of abductive explanations that are more general than those derivable without taking such rules into account. Experiments show that for the two tree-based models under consideration, benefits can be offered by the approach in terms of predictive performance and in terms of explanation sizes.


A Rectification-Based Approach for Distilling Boosted Trees into Decision Trees

Audemard, Gilles, Coste-Marquis, Sylvie, Marquis, Pierre, Sabiri, Mehdi, Szczepanski, Nicolas

arXiv.org Artificial Intelligence

We present a new approach for distilling boosted trees into decision trees, in the objective of generating an ML model offering an acceptable compromise in terms of predictive performance and interpretability. We explain how the correction approach called rectification can be used to implement such a distillation process. We show empirically that this approach provides interesting results, in comparison with an approach to distillation achieved by retraining the model.


Robust Minimax Boosting with Performance Guarantees

Mazuelas, Santiago, Alvarez, Veronica

arXiv.org Machine Learning

Boosting methods often achieve excellent classification accuracy, but can experience notable performance degradation in the presence of label noise. Existing robust methods for boosting provide theoretical robustness guarantees for certain types of label noise, and can exhibit only moderate performance degradation. However, previous theoretical results do not account for realistic types of noise and finite training sizes, and existing robust methods can provide unsatisfactory accuracies, even without noise. This paper presents methods for robust minimax boosting (RMBoost) that minimize worst-case error probabilities and are robust to general types of label noise. In addition, we provide finite-sample performance guarantees for RMBoost with respect to the error obtained without noise and with respect to the best possible error (Bayes risk). The experimental results corroborate that RMBoost is not only resilient to label noise but can also provide strong classification accuracy.